Structure from Local Optima: Learning Subspace Juntas via Higher Order PCA

نویسندگان

  • Santosh Vempala
  • Ying Xiao
چکیده

Independent Component Analysis (ICA), a well-known approach in statistics, assumes that data is generated by applying an affine transformation of a fully independent set of random variables, and aims to recover the orthogonal basis corresponding to the independent random variables. We consider a generalization of ICA, wherein the data is generated as an affine transformation applied to a product of distributions on two orthogonal subspaces, and the goal is to recover the two component subspaces. Our main result, extending the work of Frieze, Jerrum and Kannan, is an algorithm for generalized ICA that uses local optima of high moments and recovers the component subspaces. When one component is on a k-dimensional " relevant " subspace and satisfies some mild assumptions while the other is " noise " modeled as an (n − k)-dimensional Gaussian, the complexity of the algorithm is T (k, ǫ) + poly(n) where T depends only on the k-dimensional distribution. We apply this result to learning a k-subspace junta, i.e., an unknown 0-1 function in R n determined by an unknown k-dimensional subspace. This is a common generalization of learning a k-junta in R n and of learning an intersection of k halfspaces in R n , two important problems in learning theory. Our main tools are the use of local optima to recover global structure, a gradient-based algorithm for optimization over tensors, and an approximate polynomial identity test. Together, they significantly extend ICA and the class of k-dimensional labeling functions that can be learned efficiently.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Image alignment via kernelized feature learning

Machine learning is an application of artificial intelligence that is able to automatically learn and improve from experience without being explicitly programmed. The primary assumption for most of the machine learning algorithms is that the training set (source domain) and the test set (target domain) follow from the same probability distribution. However, in most of the real-world application...

متن کامل

Structure Learning of Markov Logic Networks through Iterated Local Search

Many real-world applications of AI require both probability and first-order logic to deal with uncertainty and structural complexity. Logical AI has focused mainly on handling complexity, and statistical AI on handling uncertainty. Markov Logic Networks (MLNs) are a powerful representation that combine Markov Networks (MNs) and first-order logic by attaching weights to first-order formulas and ...

متن کامل

PCA-guided search for K-means

K-means is undoubtedly themostwidely used partitional clustering algorithm. Unfortunately, due to the nonconvexity of the model formulations, expectation-maximization (EM) type algorithms converge to different local optima with different initializations. Recent discoveries have identified that the global solution of K-means cluster centroids lies in the principal component analysis (PCA) subspa...

متن کامل

A Differential Evolution and Spatial Distribution based Local Search for Training Fuzzy Wavelet Neural Network

Abstract   Many parameter-tuning algorithms have been proposed for training Fuzzy Wavelet Neural Networks (FWNNs). Absence of appropriate structure, convergence to local optima and low speed in learning algorithms are deficiencies of FWNNs in previous studies. In this paper, a Memetic Algorithm (MA) is introduced to train FWNN for addressing aforementioned learning lacks. Differential Evolution...

متن کامل

Motion Segmentation via Global and Local Sparse Subspace Optimization

In this paper, we propose a new framework for segmenting feature-based moving objects under affine subspace model. Since the feature trajectories in practice are highdimensional and contain a lot of noise, we firstly apply the sparse PCA to represent the original trajectories with a lowdimensional global subspace, which consists of the orthogonal sparse principal vectors. Subsequently, the loca...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1108.3329  شماره 

صفحات  -

تاریخ انتشار 2011